Overcoming Catastrophic Forgetting via Direction-Constrained Optimization
نویسندگان
چکیده
This paper studies a new design of the optimization algorithm for training deep learning models with fixed architecture classification network in continual framework. The data is non-stationary and non-stationarity imposed by sequence distinct tasks. We first analyze model trained on only one task isolation identify region parameter space, where performance close to recovered optimum. provide empirical evidence that this resembles cone expands along convergence direction. study principal directions trajectory optimizer after show traveling few top can quickly bring parameters outside but not case remaining directions. argue catastrophic forgetting setting be alleviated when are constrained stay within intersection plausible cones individual tasks were so far encountered during training. Based observation we present our direction-constrained (DCO) method, each introduce linear autoencoder approximate its corresponding forbidden They then incorporated into loss function form regularization term purpose coming without forgetting. Furthermore, order control memory growth as number increases, propose memory-efficient version called compressed DCO (DCO-COMP) allocates size storing all autoencoders. empirically demonstrate performs favorably compared other state-of-art regularization-based methods. codes publicly available at https://github.com/yunfei-teng/DCO .
منابع مشابه
Overcoming Catastrophic Forgetting by Incremental Moment Matching
Catastrophic forgetting is a problem of neural networks that loses the information of the first task after training the second task. Here, we propose a method, i.e. incremental moment matching (IMM), to resolve this problem. IMM incrementally matches the moment of the posterior distribution of the neural network which is trained on the first and the second task, respectively. To make the search...
متن کاملOvercoming catastrophic forgetting in neural networks
The ability to learn tasks in a sequential fashion is crucial to the development of artificial intelligence. Until now neural networks have not been capable of this and it has been widely thought that catastrophic forgetting is an inevitable feature of connectionist models. We show that it is possible to overcome this limitation and train networks that can maintain expertise on tasks that they ...
متن کاملOvercoming catastrophic forgetting with hard attention to the task
Catastrophic forgetting occurs when a neural network loses the information learned in a previous task after training on subsequent tasks. This problem remains a hurdle for artificial intelligence systems with sequential learning capabilities. In this paper, we propose a task-based hard attention mechanism that preserves previous tasks’ information without affecting the current task’s learning. ...
متن کاملAlleviating Catastrophic Forgetting via Multi-Objective Learning [IJCNN1762]
Handling catastrophic forgetting is an interesting and challenging topic in modeling the memory mechanisms of the human brain using machine learning models. From a more general point of view, catastrophic forgetting reflects the stability-plasticity dilemma, which is one of the several dilemmas to be addressed in learning systems: to retain the stored memory while learning new information. Diff...
متن کاملCatastrophic forgetting in connectionist networks.
All natural cognitive systems, and, in particular, our own, gradually forget previously learned information. Plausible models of human cognition should therefore exhibit similar patterns of gradual forgetting of old information as new information is acquired. Only rarely does new learning in natural cognitive systems completely disrupt or erase previously learned information; that is, natural c...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Lecture Notes in Computer Science
سال: 2023
ISSN: ['1611-3349', '0302-9743']
DOI: https://doi.org/10.1007/978-3-031-26387-3_41